1,550 research outputs found

    Deep Learning of Geometric Constellation Shaping including Fiber Nonlinearities

    Get PDF
    A new geometric shaping method is proposed, leveraging unsupervised machine learning to optimize the constellation design. The learned constellation mitigates nonlinear effects with gains up to 0.13 bit/4D when trained with a simplified fiber channel model.Comment: 3 pages, 6 figures, submitted to ECOC 201

    Inverse association between diabetes and altitude: a cross-sectional study in the adult population of the United States.

    Get PDF
    ObjectiveTo determine whether geographical elevation is inversely associated with diabetes, while adjusting for multiple risk factors.MethodsThis is a cross-sectional analysis of publicly available online data from the Behavioral Risk Factor Surveillance System, 2009. Final dataset included 285,196 US adult subjects. Odds ratios were obtained from multilevel mixed-effects logistic regression analysis.ResultsAmong US adults (≥20 years old), the odds ratio for diabetes was 1.00 between 0 and 499 m of altitude (reference), 0.95 (95% confidence interval, 0.90-1.01) between 500 and 1,499 m, and 0.88 (0.81-0.96) between 1,500 and 3,500 m, adjusting for age, sex, body mass index, ethnicity, self-reported fruit and vegetable consumption, self-reported physical activity, current smoking status, level of education, income, health status, employment status, and county-level information on migration rate, urbanization, and latitude. The inverse association between altitude and diabetes in the US was found among men [0.84 (0.76-0.94)], but not women [1.09 (0.97-1.22)].ConclusionsAmong US adults, living at high altitude (1,500-3,500 m) is associated with lower odds of having diabetes than living between 0 and 499 m, while adjusting for multiple risk factors. Our findings suggest that geographical elevation may be an important factor linked to diabetes

    TinyReptile: TinyML with Federated Meta-Learning

    Full text link
    Tiny machine learning (TinyML) is a rapidly growing field aiming to democratize machine learning (ML) for resource-constrained microcontrollers (MCUs). Given the pervasiveness of these tiny devices, it is inherent to ask whether TinyML applications can benefit from aggregating their knowledge. Federated learning (FL) enables decentralized agents to jointly learn a global model without sharing sensitive local data. However, a common global model may not work for all devices due to the complexity of the actual deployment environment and the heterogeneity of the data available on each device. In addition, the deployment of TinyML hardware has significant computational and communication constraints, which traditional ML fails to address. Considering these challenges, we propose TinyReptile, a simple but efficient algorithm inspired by meta-learning and online learning, to collaboratively learn a solid initialization for a neural network (NN) across tiny devices that can be quickly adapted to a new device with respect to its data. We demonstrate TinyReptile on Raspberry Pi 4 and Cortex-M4 MCU with only 256-KB RAM. The evaluations on various TinyML use cases confirm a resource reduction and training time saving by at least two factors compared with baseline algorithms with comparable performance.Comment: Accepted by The International Joint Conference on Neural Network (IJCNN) 202

    Statistics Oriented Preprocessing of Document Image

    Get PDF
    Old printed documents represent an important part of our cultural heritage. Their digitalization plays an important role in creating data and metadata. The paper proposed an algorithm for estimation of the global text skew. First, document image is binarized reducing the impact of noise and uneven illumination. The binary image is statistically analyzed and processed. Accordingly, redundant data have been excluded. Furthermore, the convex hulls are established encircling each text object. They are joined establishing connected components. Then, the connected components in complementary image are enlarged with morphological dilation. At the end, the biggest connected component is extracted. Its orientation is similar to the global orientation of text document which is calculated by the moments. Efficiency and correctness of the algorithm are verified by testing on a custom dataset

    Main memory in HPC: do we need more, or could we live with less?

    Get PDF
    An important aspect of High-Performance Computing (HPC) system design is the choice of main memory capacity. This choice becomes increasingly important now that 3D-stacked memories are entering the market. Compared with conventional Dual In-line Memory Modules (DIMMs), 3D memory chiplets provide better performance and energy efficiency but lower memory capacities. Therefore, the adoption of 3D-stacked memories in the HPC domain depends on whether we can find use cases that require much less memory than is available now. This study analyzes the memory capacity requirements of important HPC benchmarks and applications. We find that the High-Performance Conjugate Gradients (HPCG) benchmark could be an important success story for 3D-stacked memories in HPC, but High-Performance Linpack (HPL) is likely to be constrained by 3D memory capacity. The study also emphasizes that the analysis of memory footprints of production HPC applications is complex and that it requires an understanding of application scalability and target category, i.e., whether the users target capability or capacity computing. The results show that most of the HPC applications under study have per-core memory footprints in the range of hundreds of megabytes, but we also detect applications and use cases that require gigabytes per core. Overall, the study identifies the HPC applications and use cases with memory footprints that could be provided by 3D-stacked memory chiplets, making a first step toward adoption of this novel technology in the HPC domain.This work was supported by the Collaboration Agreement between Samsung Electronics Co., Ltd. and BSC, Spanish Government through Severo Ochoa programme (SEV-2015-0493), by the Spanish Ministry of Science and Technology through TIN2015-65316-P project and by the Generalitat de Catalunya (contracts 2014-SGR-1051 and 2014-SGR-1272). This work has also received funding from the European Union’s Horizon 2020 research and innovation programme under ExaNoDe project (grant agreement No 671578). Darko Zivanovic holds the Severo Ochoa grant (SVP-2014-068501) of the Ministry of Economy and Competitiveness of Spain. The authors thank Harald Servat from BSC and Vladimir Marjanovi´c from High Performance Computing Center Stuttgart for their technical support.Postprint (published version
    corecore